Sobolev gradients of viscosity supersolutions
نویسندگان
چکیده
We investigate which elliptic PDEs have the property that every viscosity supersolution belongs to a Sobolev space $$W^{1,q}_{loc}(\Omega )$$ , $$\Omega \subseteq \mathbb {R}^n$$ . The asymptotic cone of operator’s sublevel set seems be essential. It turns out much can said if we know how is related dominative p-Laplacian, with $$p = \frac{n-1}{n}q + 1$$ show that, in certain sense, this minimal operator associated exponent q.
منابع مشابه
VISCOSITY SUPERSOLUTIONS OF THE EVOLUTIONARY p-LAPLACE EQUATION
has this character. Even obvious results for this equation may require advanced estimates in the proofs. We refer to the books [DB] and [WZYL] about this equation, which is called the “evolutionary p-Laplacian equation,” the “p-parabolic equation” or even the “non-Newtonian equation of filtration.”. Our objective is to study the regularity of the viscosity supersolutions and their spatial gradi...
متن کاملSobolev Gradients: Introduction, Applications, Problems
Sobolev gradient is defined and a simple example is given. A number of applications are described. A number of problems are stated, some of which are open problems for research.
متن کاملSobolev Gradients and the Ginzburg-Landau Functional
We describe a Sobolev gradient method for finding minima of the Ginzburg–Landau functional for superconductivity. This method leads to a particularly simple algorithm which avoids consideration of the nonlinear boundary conditions associated with the Ginzburg–Landau equations.
متن کاملGeometric Curve Modeling with Sobolev Gradients
The Sobolev gradient method is a powerful tool for geometric modeling. We treat the problem of constructing fair curves by minimizing a fairness measure subject to geometric constraints. The measure might include curve length, curvature, torsion, and/or variation of curvature. The constraints may include specified values, tangent vectors, and/or curvature vectors. We may also require periodicit...
متن کاملNonlinear least squares and Sobolev gradients
Least squares methods are effective for solving systems of partial differential equations. In the case of nonlinear systems the equations are usually linearized by a Newton iteration or successive substitution method, and then treated as a linear least squares problem. We show that it is often advantageous to form a sum of squared residuals first, and then compute a zero of the gradient with a ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Nonlinear Differential Equations And Applications Nodea
سال: 2023
ISSN: ['1420-9004', '1021-9722']
DOI: https://doi.org/10.1007/s00030-023-00858-8